Search results for " Reinforcement learning"

showing 10 items of 13 documents

Using Inverse Reinforcement Learning with Real Trajectories to Get More Trustworthy Pedestrian Simulations

2020

Reinforcement learning is one of the most promising machine learning techniques to get intelligent behaviors for embodied agents in simulations. The output of the classic Temporal Difference family of Reinforcement Learning algorithms adopts the form of a value function expressed as a numeric table or a function approximator. The learned behavior is then derived using a greedy policy with respect to this value function. Nevertheless, sometimes the learned policy does not meet expectations, and the task of authoring is difficult and unsafe because the modification of one value or parameter in the learned value function has unpredictable consequences in the space of the policies it represents…

0209 industrial biotechnologyreinforcement learningComputer scienceGeneral Mathematics02 engineering and technologypedestrian simulationTask (project management)learning by demonstration020901 industrial engineering & automationAprenentatgeInformàticaBellman equation0202 electrical engineering electronic engineering information engineeringComputer Science (miscellaneous)Reinforcement learningEngineering (miscellaneous)business.industrycausal entropylcsh:MathematicsProcess (computing)020206 networking & telecommunicationsFunction (mathematics)inverse reinforcement learninglcsh:QA1-939Problem domainTable (database)Artificial intelligenceTemporal difference learningbusinessoptimizationMathematics
researchProduct

Emergent behaviors and scalability for multi-agent reinforcement learning-based pedestrian models

2017

This paper analyzes the emergent behaviors of pedestrian groups that learn through the multiagent reinforcement learning model developed in our group. Five scenarios studied in the pedestrian model literature, and with different levels of complexity, were simulated in order to analyze the robustness and the scalability of the model. Firstly, a reduced group of agents must learn by interaction with the environment in each scenario. In this phase, each agent learns its own kinematic controller, that will drive it at a simulation time. Secondly, the number of simulated agents is increased, in each scenario where agents have previously learnt, to test the appearance of emergent macroscopic beha…

Engineeringmedia_common.quotation_subject02 engineering and technologyPedestrianMachine learningcomputer.software_genreConsistency (database systems)Robustness (computer science)0202 electrical engineering electronic engineering information engineeringReinforcement learningQuality (business)Macromedia_commonInformáticaPedestrian simulation and modelingKinematic controllerbusiness.industry020207 software engineeringEmergent behavioursBehavioural simulationHardware and ArchitectureModeling and SimulationScalability020201 artificial intelligence & image processingArtificial intelligencebusinessMulti-agent reinforcement learning (Marl)computerSoftwareSimulation Modelling Practice and Theory
researchProduct

Robust Adaptive Modulation and Coding (AMC) selection in LTE systems using reinforcement learning

2014

Adaptive Modulation and Coding (AMC) in LTE networks is commonly employed to improve system throughput by ensuring more reliable transmissions. Most of existing AMC methods select the modulation and coding scheme (MCS) using pre-computed mappings between MCS indexes and channel quality indicator (CQI) feedbacks that are periodically sent by the receivers. However, the effectiveness of this approach heavily depends on the assumed channel model. In addition CQI feedback delays may cause throughput losses. In this paper we design a new AMC scheme that exploits a reinforcement learning algorithm to adjust at run-time the MCS selection rules based on the knowledge of the effect of previous AMC d…

Engineeringreinforcement learningSettore ING-INF/03 - Telecomunicazionibusiness.industryLink adaptationchannel qualityChannel modelsLTE channel quality adaptive modulation and coding (AMC) reinforcement learning performance evaluation.performance evaluationLTERobustness (computer science)Electronic engineeringReinforcement learningDecision processbusinessReinforcement learning algorithmCoding (social sciences)adaptive modulation and coding (AMC)
researchProduct

Can Interpretable Reinforcement Learning Manage Prosperity Your Way?

2022

Personalisation of products and services is fast becoming the driver of success in banking and commerce. Machine learning holds the promise of gaining a deeper understanding of and tailoring to customers’ needs and preferences. Whereas traditional solutions to financial decision problems frequently rely on model assumptions, reinforcement learning is able to exploit large amounts of data to improve customer modelling and decision-making in complex financial environments with fewer assumptions. Model explainability and interpretability present challenges from a regulatory perspective which demands transparency for acceptance; they also offer the opportunity for improved insight into and unde…

FOS: Computer and information sciencesComputer Science - Machine LearningArtificial Intelligence (cs.AI)Computer Science - Artificial IntelligenceGeneral Earth and Planetary SciencesAI in banking; personalized services; prosperity management; explainable AI; reinforcement learning; policy regularisationVDP::Teknologi: 500::Informasjons- og kommunikasjonsteknologi: 550General Environmental ScienceMachine Learning (cs.LG)AI; Volume 3; Issue 2; Pages: 526-537
researchProduct

Expanding the Active Inference Landscape: More Intrinsic Motivations in the Perception-Action Loop

2018

Active inference is an ambitious theory that treats perception, inference and action selection of autonomous agents under the heading of a single principle. It suggests biologically plausible explanations for many cognitive phenomena, including consciousness. In active inference, action selection is driven by an objective function that evaluates possible future actions with respect to current, inferred beliefs about the world. Active inference at its core is independent from extrinsic rewards, resulting in a high level of robustness across e.g.\ different environments or agent morphologies. In the literature, paradigms that share this independence have been summarised under the notion of in…

FOS: Computer and information sciencesComputer scienceComputer Science - Artificial Intelligencepredictive informationBiomedical EngineeringInferenceSystems and Control (eess.SY)02 engineering and technologyAction selectionI.2.0; I.2.6; I.5.0; I.5.1lcsh:RC321-57103 medical and health sciences0302 clinical medicineactive inferenceArtificial IntelligenceFOS: Electrical engineering electronic engineering information engineering0202 electrical engineering electronic engineering information engineeringFormal concept analysisMethodsperception-action loopuniversal reinforcement learningintrinsic motivationlcsh:Neurosciences. Biological psychiatry. NeuropsychiatryFree energy principleCognitive scienceRobotics and AII.5.0I.5.1I.2.6Partially observable Markov decision processI.2.0Artificial Intelligence (cs.AI)Action (philosophy)empowermentIndependence (mathematical logic)free energy principleComputer Science - Systems and Control020201 artificial intelligence & image processingBiological plausibility62F15 91B06030217 neurology & neurosurgeryvariational inference
researchProduct

On the use of Deep Reinforcement Learning for Visual Tracking: a Survey

2021

This paper aims at highlighting cutting-edge research results in the field of visual tracking by deep reinforcement learning. Deep reinforcement learning (DRL) is an emerging area combining recent progress in deep and reinforcement learning. It is showing interesting results in the computer vision field and, recently, it has been applied to the visual tracking problem yielding to the rapid development of novel tracking strategies. After providing an introduction to reinforcement learning, this paper compares recent visual tracking approaches based on deep reinforcement learning. Analysis of the state-of-the-art suggests that reinforcement learning allows modeling varying parts of the tracki…

General Computer ScienceComputer scienceFeature extractionMachine learningcomputer.software_genreField (computer science)video-surveillanceMinimum bounding boxReinforcement learningGeneral Materials ScienceSettore ING-INF/05 - Sistemi Di Elaborazione Delle Informazionideep reinforcement learningComputer vision machine learning video-surveillance deep reinforcement learning visual tracking.business.industryGeneral EngineeringTracking systemvisual trackingVisualizationActive appearance modelTK1-9971machine learningEye trackingComputer visionArtificial intelligenceElectrical engineering. Electronics. Nuclear engineeringbusinesscomputer
researchProduct

Development of a Simulator for Prototyping Reinforcement Learning-Based Autonomous Cars

2022

Autonomous driving is a research field that has received attention in recent years, with increasing applications of reinforcement learning (RL) algorithms. It is impractical to train an autonomous vehicle thoroughly in the physical space, i.e., the so-called ’real world’; therefore, simulators are used in almost all training of autonomous driving algorithms. There are numerous autonomous driving simulators, very few of which are specifically targeted at RL. RL-based cars are challenging due to the variety of reward functions available. There is a lack of simulators addressing many central RL research tasks within autonomous driving, such as scene understanding, localization and mapping, pla…

Human-Computer InteractionVDP::Teknologi: 500autonomous driving; simulators; reinforcement learningComputer Networks and CommunicationsCommunicationInformatics
researchProduct

Designing a multi-layer edge-computing platform for energy-efficient and delay-aware offloading in vehicular networks

2021

Abstract Vehicular networks are expected to support many time-critical services requiring huge amounts of computation resources with very low delay. However, such requirements may not be fully met by vehicle on-board devices due to their limited processing and storage capabilities. The solution provided by 5G is the application of the Multi-Access Edge Computing (MEC) paradigm, which represents a low-latency alternative to remote clouds. Accordingly, we envision a multi-layer job-offloading scheme based on three levels, i.e., the Vehicular Domain, the MEC Domain and Backhaul Network Domain. In such a view, jobs can be offloaded from the Vehicular Domain to the MEC Domain, and even further o…

Markov ModelsVehicular ad hoc networkComputer Networks and CommunicationsComputer scienceDistributed computing5G; Edge Computing; Markov Models; Reinforcement Learning; Vehicular NetworksLoad balancing (computing)Reinforcement LearningDomain (software engineering)ServerEdge ComputingReinforcement learningVehicular NetworksMarkov decision process5GEdge computingEfficient energy useComputer Networks
researchProduct

OMNI-DRL: Learning to Fly in Forests with Omnidirectional Images

2022

Perception is crucial for drone obstacle avoidance in complex, static, and unstructured outdoor environments. However, most navigation solutions based on Deep Reinforcement Learning (DRL) use limited Field-Of-View (FOV) images as input. In this paper, we demonstrate that omnidirectional images improve these methods. Thus, we provide a comparative benchmark of several visual modalities for navigation: ground truth depth, ground truth semantic segmentation, and RGB images. These exhaustive comparisons reveal that it is superior to use an omnidirectional camera to navigate with classical DRL methods. Finally, we show in two different virtual forest environments that adapting the convolution to…

Perception and sensingDeep Reinforcement LearningControl and Systems EngineeringMobile robots and vehicles[INFO.INFO-RB] Computer Science [cs]/Robotics [cs.RO]Omnidirectional sensorsLearning robot control
researchProduct

Roboception and adaptation in a cognitive robot

2023

In robotics, perception is usually oriented at understanding what is happening in the external world, while few works pay attention to what is occurring in the robot’s body. In this work, we propose an artificial somatosensory system, embedded in a cognitive architecture, that enables a robot to perceive the sensations from its embodiment while executing a task. We called these perceptions roboceptions, and they let the robot act according to its own physical needs in addition to the task demands. Physical information is processed by the robot to behave in a balanced way, determining the most appropriate trade-off between the achievement of the task and its well being. The experiments show …

Settore ING-INF/05 - Sistemi Di Elaborazione Delle InformazioniControl and Systems EngineeringGeneral MathematicsSomatosensory system Soft sensors Cognitive architecture Humanoid robot Roboception Reinforcement learningSoftwareComputer Science Applications
researchProduct